Deep Learning with Apache SystemML

نویسندگان

  • Niketan Pansare
  • Michael Dusenberry
  • Nakul Jindal
  • Matthias Boehm
  • Berthold Reinwald
  • Prithviraj Sen
چکیده

Deep Learning (DL) is a subfield of Machine Learning (ML) that focuses on learning hierarchical representations of data with multiple levels of abstraction using neural networks [15]. Recent advances in deep learning are made possible due to the availability of large amounts of labeled data, use of GPGPU compute, and application of new techniques (such as ReLU, batch normalization [12], dropout [17], residual block [10], etc.) that help deal with issues in training deep networks. In spite of the need to train on large datasets, there is a disconnect between the deep learning community and the big data community. To scale to a multi-node cluster, most deep learning frameworks (such as Caffe2, TensorFlow [2] and IBM’s PowerAI DDL [6]) use custom communication libraries based on either MPI (such as IBM SpectrumMPI, Facebook’s Gloo) or a customnetworking protocol (such as Google RPC). Unlike popular big data frameworks (such as Apache Hadoop [9] and Apache Spark [18]), these communication libraries do not provide features such as resource sharing, multi-tenancy and fault-tolerance out of the box, making them difficult to deploy on shared production clusters. Œis leads to ineffective use of resources in an organization, o‰en requiring two separate infrastructures (i.e. scale-up versus scale-out). Œis problem is even more severe when the data generated as part of the big data pipeline (ML, data preprocessing, data cleaning) needs to be consumed by the deep learning pipeline or vice versa, as the workload characteristics of a typical machine learning algorithm (i.e. memory-bound, BLAS level-2, sparse/ultrasparse inputs (or feature matrix), etc.) are o‰en different than that of a typical deep learning algorithm (i.e. compute-bound, BLAS level-3, dense inputs, etc.). Apache SystemML [4] aims to bridge that gap by seamlessly integrating with underlying big data frameworks and by providing a unified framework for implementing machine learning and deep learning algorithms. In Apache SystemML, the ML algorithms are implemented using a high-level R-like language called DML (short for Declarative Machine Learning). DML improves the productivity of data scientists by enabling them to implement their ML algorithm with precise semantics as well as abstract data types and operations, independent of the underlying data representation or cluster characteristics. For the given DML script, SystemML’s cost-based compiler automatically generates hybrid runtime execution plans that are composed of single-node and distributed operations depending on data and cluster characteristics such as data size, data sparsity, cluster size and memory configurations, while exploiting the capabilities of underlying data-parallel frameworks such asMapReduce or Spark. Œis allows for algorithm reusability across data-parallel frameworks, and simplified deployment for varying data characteristics and runtime environments, ranging from low-latency scoring to large-scale training. 2 DEEP LEARNING APIS

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

SystemML: Declarative Machine Learning on Spark

The rising need for custom machine learning (ML) algorithms and the growing data sizes that require the exploitation of distributed, data-parallel frameworks such as MapReduce or Spark, pose significant productivity challenges to data scientists. Apache SystemML addresses these challenges through declarative ML by (1) increasing the productivity of data scientists as they are able to express cu...

متن کامل

On Optimizing Operator Fusion Plans for Large-Scale Machine Learning in SystemML

Many large-scale machine learning (ML) systems allow specifying custom ML algorithms by means of linear algebra programs, and then automatically generate efficient execution plans. In this context, optimization opportunities for fused operators—in terms of fused chains of basic operators—are ubiquitous. These opportunities include (1) fewer materialized intermediates, (2) fewer scans of input d...

متن کامل

Hybrid Parallelization Strategies for Large-Scale Machine Learning in SystemML

SystemML aims at declarative, large-scale machine learning (ML) on top of MapReduce, where high-level ML scripts with R-like syntax are compiled to programs of MR jobs. The declarative specification of ML algorithms enables—in contrast to existing large-scale machine learning libraries— automatic optimization. SystemML’s primary focus is on data parallelism but many ML algorithms inherently exh...

متن کامل

SystemML's Optimizer: Plan Generation for Large-Scale Machine Learning Programs

SystemML enables declarative, large-scale machine learning (ML) via a high-level language with R-like syntax. Data scientists use this language to express their ML algorithms with full flexibility but without the need to hand-tune distributed runtime execution plans and system configurations. These ML programs are dynamically compiled and optimized based on data and cluster characteristics usin...

متن کامل

Deep Learning-Based Approach for Time Series Forecasting with Application to Electricity Load

This paper presents a novel method to predict times series using deep learning. In particular, the method can be used for arbitrary time horizons, dividing each predicted sample into a single problem. This fact allows easy parallelization and adaptation to the big data context. Deep learning implementation in H2O library is used for each subproblem. However, H2O does not permit multi-step regre...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1802.04647  شماره 

صفحات  -

تاریخ انتشار 2018